cxr dataset
Improving Pediatric Pneumonia Diagnosis with Adult Chest X-ray Images Utilizing Contrastive Learning and Embedding Similarity
Zunaed, Mohammad, Hasan, Anwarul, Hasan, Taufiq
Despite the advancement of deep learning-based computer-aided diagnosis (CAD) methods for pneumonia from adult chest x-ray (CXR) images, the performance of CAD methods applied to pediatric images remains suboptimal, mainly due to the lack of large-scale annotated pediatric imaging datasets. Establishing a proper framework to leverage existing adult large-scale CXR datasets can thus enhance pediatric pneumonia detection performance. In this paper, we propose a three-branch parallel path learning-based framework that utilizes both adult and pediatric datasets to improve the performance of deep learning models on pediatric test datasets. The paths are trained with pediatric only, adult only, and both types of CXRs, respectively. Our proposed framework utilizes the multi-positive contrastive loss to cluster the classwise embeddings and the embedding similarity loss among these three parallel paths to make the classwise embeddings as close as possible to reduce the effect of domain shift. Experimental evaluations on open-access adult and pediatric CXR datasets show that the proposed method achieves a superior AUROC score of 0.8464 compared to 0.8348 obtained using the conventional approach of join training on both datasets. The proposed approach thus paves the way for generalized CAD models that are effective for both adult and pediatric age groups.
- North America > United States > Maryland > Baltimore (0.04)
- Asia > Middle East > Qatar > Ad-Dawhah > Doha (0.04)
- Asia > Bangladesh > Dhaka Division > Dhaka District > Dhaka (0.04)
Surgical Aggregation: Federated Class-Heterogeneous Learning
Kulkarni, Pranav, Kanhere, Adway, Yi, Paul H., Parekh, Vishwa S.
Abstract-- The release of numerous chest x-ray datasets has spearheaded the development of deep learning models with expert-level performance. However, they have limited interoperability due to class-heterogeneity - a result of inconsistent labeling schemes and partial annotations. Therefore, it is challenging to leverage these datasets in aggregate to train models with a complete representation of abnormalities that may occur within the thorax. In this work, we propose surgical aggregation, a federated learning framework for aggregating knowledge from class-heterogeneous datasets and learn a model that can simultaneously predict the presence of all disease labels present across the datasets. We evaluate our method using simulated and real-world class-heterogeneous datasets across both independent and identically distributed (iid) and non-iid settings. Our results show that surgical aggregation outperforms current methods, has better generalizability, and is a crucial first step towards tackling class-heterogeneity in federated learning to facilitate the development of clinically-useful models using previously non-interoperable chest x-ray datasets.
- South America > Peru > Lima Department > Lima Province > Lima (0.04)
- North America > United States > Maryland > Baltimore (0.04)
BoMD: Bag of Multi-label Descriptors for Noisy Chest X-ray Classification
Chen, Yuanhong, Liu, Fengbei, Wang, Hu, Wang, Chong, Tian, Yu, Liu, Yuyuan, Carneiro, Gustavo
Deep learning methods have shown outstanding classification accuracy in medical imaging problems, which is largely attributed to the availability of large-scale datasets manually annotated with clean labels. However, given the high cost of such manual annotation, new medical imaging classification problems may need to rely on machine-generated noisy labels extracted from radiology reports. Indeed, many Chest X-ray (CXR) classifiers have already been modelled from datasets with noisy labels, but their training procedure is in general not robust to noisy-label samples, leading to sub-optimal models. Furthermore, CXR datasets are mostly multi-label, so current noisy-label learning methods designed for multi-class problems cannot be easily adapted. In this paper, we propose a new method designed for the noisy multi-label CXR learning, which detects and smoothly re-labels samples from the dataset, which is then used to train common multi-label classifiers. The proposed method optimises a bag of multi-label descriptors (BoMD) to promote their similarity with the semantic descriptors produced by BERT models from the multi-label image annotation. Our experiments on diverse noisy multi-label training sets and clean testing sets show that our model has state-of-the-art accuracy and robustness in many CXR multi-label classification benchmarks.
- Health & Medicine > Nuclear Medicine (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (1.00)
From Competition to Collaboration: Making Toy Datasets on Kaggle Clinically Useful for Chest X-Ray Diagnosis Using Federated Learning
Kulkarni, Pranav, Kanhere, Adway, Yi, Paul H., Parekh, Vishwa S.
Chest X-ray (CXR) datasets hosted on Kaggle, though useful from a data science competition standpoint, have limited utility in clinical use because of their narrow focus on diagnosing one specific disease. In real-world clinical use, multiple diseases need to be considered since they can co-exist in the same patient. In this work, we demonstrate how federated learning (FL) can be used to make these toy CXR datasets from Kaggle clinically useful. Specifically, we train a single FL classification model (`global`) using two separate CXR datasets -- one annotated for presence of pneumonia and the other for presence of pneumothorax (two common and life-threatening conditions) -- capable of diagnosing both. We compare the performance of the global FL model with models trained separately on both datasets (`baseline`) for two different model architectures. On a standard, naive 3-layer CNN architecture, the global FL model achieved AUROC of 0.84 and 0.81 for pneumonia and pneumothorax, respectively, compared to 0.85 and 0.82, respectively, for both baseline models (p>0.05). Similarly, on a pretrained DenseNet121 architecture, the global FL model achieved AUROC of 0.88 and 0.91 for pneumonia and pneumothorax, respectively, compared to 0.89 and 0.91, respectively, for both baseline models (p>0.05). Our results suggest that FL can be used to create global `meta` models to make toy datasets from Kaggle clinically useful, a step forward towards bridging the gap from bench to bedside.
- Health & Medicine > Nuclear Medicine (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (1.00)
CheXclusion: Fairness gaps in deep chest X-ray classifiers
Seyyed-Kalantari, Laleh, Liu, Guanxiong, McDermott, Matthew, Ghassemi, Marzyeh
Machine learning systems have received much attention recently for their ability to achieve expert-level performance on clinical tasks, particularly in medical imaging. Here, we examine the extent to which state-of-the-art deep learning classifiers trained to yield diagnostic labels from X-ray images are biased with respect to protected attributes. We train convolution neural networks to predict 14 diagnostic labels in three prominent public chest X-ray datasets: MIMIC-CXR, Chest-Xray8, and CheXpert. We then evaluate the TPR disparity - the difference in true positive rates (TPR) and - underdiagnosis rate - the false positive rate of a non-diagnosis - among different protected attributes such as patient sex, age, race, and insurance type. We demonstrate that TPR disparities exist in the state-of-the-art classifiers in all datasets, for all clinical tasks, and all subgroups. We find that TPR disparities are most commonly not significantly correlated with a subgroup's proportional disease burden; further, we find that some subgroups and subsection of the population are chronically underdiagnosed. Such performance disparities have real consequences as models move from papers to products, and should be carefully audited prior to deployment.
- North America > Canada > Ontario > Toronto (0.14)
- North America > United States > Georgia > Fulton County > Atlanta (0.04)
- North America > United States > Massachusetts > Middlesex County > Malden (0.04)
- (5 more...)
- Health & Medicine > Therapeutic Area (1.00)
- Health & Medicine > Nuclear Medicine (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (1.00)
Multi-Adversarial Variational Autoencoder Networks
Imran, Abdullah-Al-Zubaer, Terzopoulos, Demetri
The unsupervised training of GANs and VAEs has enabled them to generate realistic images mimicking real-world distributions and perform image-based unsupervised clustering or semi-supervised classification. Combining the power of these two generative models, we introduce Multi-Adversarial Variational autoEncoder Networks (MAVENs), a novel network architecture that incorporates an ensemble of discriminators in a VAE-GAN network, with simultaneous adversarial learning and variational inference. We apply MAVENs to the generation of synthetic images and propose a new distribution measure to quantify the quality of the generated images. Our experimental results using datasets from the computer vision and medical imaging domains---Street View House Numbers, CIFAR-10, and Chest X-Ray datasets---demonstrate competitive performance against state-of-the-art semi-supervised models both in image generation and classification tasks.
Multi Layered-Parallel Graph Convolutional Network (ML-PGCN) for Disease Prediction
Kazi, Anees, Albarqouni, Shadi, Kortuem, Karsten, Navab, Nassir
Structural data(age, gender, weight) from Electronic Health Records (EHRs) are exploited by Computer Aided Systems (CADS) as complementary information for disease prediction. Such systems, however, fail to weight the structural data based its relevance to the disease at hand. A model capable of evaluating the significance of every element of the structural data and performing the prediction task based on the selective and weighted procedure for elements of structural data is required. Such scheme will boost more semantic automatic disease prediction task Recently multi-modal data is processed using deep learning methods like Convolutional Neural Networks(CNNs)[9], Autoencoders[6], Modified Restricted Boltzman Machine[8] etc. These methods provide richer and discriminant feature space which helps to exploit the global complementary information from available modalities, however, fail to address the problem of unequal relevance. Structural data gives statistical information about the population as a whole. This is taken into consideration more recently using graphs, providing a more semantic way of using multi-modal data[7, 4]. These methods focus more on the association between the subjects with respect to either of the modalities and then solve the tasks such as disease prediction with features from other modalities.
- Asia > China > Guangdong Province > Shenzhen (0.05)
- Europe > Germany > North Rhine-Westphalia > Upper Bavaria > Munich (0.05)
- North America > United States > Maryland (0.04)
- Health & Medicine > Diagnostic Medicine > Imaging (0.69)
- Health & Medicine > Health Care Technology > Medical Record (0.55)